6 research outputs found
A Survey on ChatGPT: AI-Generated Contents, Challenges, and Solutions
With the widespread use of large artificial intelligence (AI) models such as
ChatGPT, AI-generated content (AIGC) has garnered increasing attention and is
leading a paradigm shift in content creation and knowledge representation. AIGC
uses generative large AI algorithms to assist or replace humans in creating
massive, high-quality, and human-like content at a faster pace and lower cost,
based on user-provided prompts. Despite the recent significant progress in
AIGC, security, privacy, ethical, and legal challenges still need to be
addressed. This paper presents an in-depth survey of working principles,
security and privacy threats, state-of-the-art solutions, and future challenges
of the AIGC paradigm. Specifically, we first explore the enabling technologies,
general architecture of AIGC, and discuss its working modes and key
characteristics. Then, we investigate the taxonomy of security and privacy
threats to AIGC and highlight the ethical and societal implications of GPT and
AIGC technologies. Furthermore, we review the state-of-the-art AIGC
watermarking approaches for regulatable AIGC paradigms regarding the AIGC model
and its produced content. Finally, we identify future challenges and open
research directions related to AIGC.Comment: 20 pages, 6 figures, 4 table
Social-Aware Clustered Federated Learning with Customized Privacy Preservation
A key feature of federated learning (FL) is to preserve the data privacy of
end users. However, there still exist potential privacy leakage in exchanging
gradients under FL. As a result, recent research often explores the
differential privacy (DP) approaches to add noises to the computing results to
address privacy concerns with low overheads, which however degrade the model
performance. In this paper, we strike the balance of data privacy and
efficiency by utilizing the pervasive social connections between users.
Specifically, we propose SCFL, a novel Social-aware Clustered Federated
Learning scheme, where mutually trusted individuals can freely form a social
cluster and aggregate their raw model updates (e.g., gradients) inside each
cluster before uploading to the cloud for global aggregation. By mixing model
updates in a social group, adversaries can only eavesdrop the social-layer
combined results, but not the privacy of individuals. We unfold the design of
SCFL in three steps. \emph{i) Stable social cluster formation. Considering
users' heterogeneous training samples and data distributions, we formulate the
optimal social cluster formation problem as a federation game and devise a fair
revenue allocation mechanism to resist free-riders. ii) Differentiated
trust-privacy mapping}. For the clusters with low mutual trust, we design a
customizable privacy preservation mechanism to adaptively sanitize
participants' model updates depending on social trust degrees. iii) Distributed
convergence}. A distributed two-sided matching algorithm is devised to attain
an optimized disjoint partition with Nash-stable convergence. Experiments on
Facebook network and MNIST/CIFAR-10 datasets validate that our SCFL can
effectively enhance learning utility, improve user payoff, and enforce
customizable privacy protection
Trade Privacy for Utility: A Learning-Based Privacy Pricing Game in Federated Learning
To prevent implicit privacy disclosure in sharing gradients among data owners
(DOs) under federated learning (FL), differential privacy (DP) and its variants
have become a common practice to offer formal privacy guarantees with low
overheads. However, individual DOs generally tend to inject larger DP noises
for stronger privacy provisions (which entails severe degradation of model
utility), while the curator (i.e., aggregation server) aims to minimize the
overall effect of added random noises for satisfactory model performance. To
address this conflicting goal, we propose a novel dynamic privacy pricing
(DyPP) game which allows DOs to sell individual privacy (by lowering the scale
of locally added DP noise) for differentiated economic compensations (offered
by the curator), thereby enhancing FL model utility. Considering
multi-dimensional information asymmetry among players (e.g., DO's data
distribution and privacy preference, and curator's maximum affordable payment)
as well as their varying private information in distinct FL tasks, it is hard
to directly attain the Nash equilibrium of the mixed-strategy DyPP game.
Alternatively, we devise a fast reinforcement learning algorithm with two
layers to quickly learn the optimal mixed noise-saving strategy of DOs and the
optimal mixed pricing strategy of the curator without prior knowledge of
players' private information. Experiments on real datasets validate the
feasibility and effectiveness of the proposed scheme in terms of faster
convergence speed and enhanced FL model utility with lower payment costs.Comment: Accepted by IEEE ICC202